Search results for "inductive inference"

showing 10 items of 14 documents

Learning with belief levels

2008

AbstractWe study learning of predicate logics formulas from “elementary facts,” i.e. from the values of the predicates in the given model. Several models of learning are considered, but most of our attention is paid to learning with belief levels. We propose an axiom system which describes what we consider to be a human scientist's natural behavior when trying to explore these elementary facts. It is proved that no such system can be complete. However we believe that our axiom system is “practically” complete. Theorems presented in the paper in some sense confirm our hypothesis.

CompletenessAxiom systemsbusiness.industryComputer Networks and CommunicationsApplied Mathematics010102 general mathematicsInductive inference02 engineering and technologyInductive reasoning01 natural sciencesBelief levelsPredicate (grammar)EpistemologyTheoretical Computer ScienceTheoryofComputation_MATHEMATICALLOGICANDFORMALLANGUAGESComputational Theory and Mathematics020204 information systems0202 electrical engineering electronic engineering information engineeringLearningArtificial intelligence0101 mathematicsbusinessAction axiomAxiomMathematicsJournal of Computer and System Sciences
researchProduct

On the inductive inference of recursive real-valued functions

1999

AbstractWe combine traditional studies of inductive inference and classical continuous mathematics to produce a study of learning real-valued functions. We consider two possible ways to model the learning by example of functions with domain and range the real numbers. The first approach considers functions as represented by computable analytic functions. The second considers arbitrary computable functions of recursive real numbers. In each case we find natural examples of learnable classes of functions and unlearnable classes of functions.

Complex-valued functionGeneral Computer ScienceReal analysisLearning theoryComputable numberInductive inference0102 computer and information sciences02 engineering and technology01 natural sciencesμ-recursive functionComputable analysisTheoretical Computer ScienceAlgebraμ operatorComputable functionReal-valued computationReal-valued function010201 computation theory & mathematics0202 electrical engineering electronic engineering information engineering020201 artificial intelligence & image processingAlgorithmComputer Science(all)MathematicsTheoretical Computer Science
researchProduct

Hierarchies of probabilistic and team FIN-learning

2001

AbstractA FIN-learning machine M receives successive values of the function f it is learning and at some moment outputs a conjecture which should be a correct index of f. FIN learning has two extensions: (1) If M flips fair coins and learns a function with certain probability p, we have FIN〈p〉-learning. (2) When n machines simultaneously try to learn the same function f and at least k of these machines output correct indices of f, we have learning by a [k,n]FIN team. Sometimes a team or a probabilistic learner can simulate another one, if their probabilities p1,p2 (or team success ratios k1/n1,k2/n2) are close enough (Daley et al., in: Valiant, Waranth (Eds.), Proc. 5th Annual Workshop on C…

Discrete mathematics020203 distributed computingProbabilistic learningConjectureFinGeneral Computer ScienceIndex (typography)Probabilistic logicInductive inference0102 computer and information sciences02 engineering and technologyFunction (mathematics)01 natural sciencesTheoretical Computer ScienceMoment (mathematics)Computational learning theory010201 computation theory & mathematics0202 electrical engineering electronic engineering information engineeringTeam learningAlgorithmComputer Science(all)MathematicsTheoretical Computer Science
researchProduct

Probabilistic and team PFIN-type learning: General properties

2008

We consider the probability hierarchy for Popperian FINite learning and study the general properties of this hierarchy. We prove that the probability hierarchy is decidable, i.e. there exists an algorithm that receives p_1 and p_2 and answers whether PFIN-type learning with the probability of success p_1 is equivalent to PFIN-type learning with the probability of success p_2. To prove our result, we analyze the topological structure of the probability hierarchy. We prove that it is well-ordered in descending ordering and order-equivalent to ordinal epsilon_0. This shows that the structure of the hierarchy is very complicated. Using similar methods, we also prove that, for PFIN-type learning…

FOS: Computer and information sciencesComputer Science::Machine LearningTheoretical computer scienceComputer Networks and CommunicationsExistential quantificationStructure (category theory)DecidabilityType (model theory)Learning in the limitTheoretical Computer ScienceMachine Learning (cs.LG)Probability of successFinite limitsMathematicsOrdinalsDiscrete mathematicsHierarchybusiness.industryApplied MathematicsAlgorithmic learning theoryProbabilistic logicF.1.1 I.2.6Inductive inferenceInductive reasoningDecidabilityComputer Science - LearningTeam learningComputational Theory and MathematicsArtificial intelligencebusinessJournal of Computer and System Sciences
researchProduct

Quantum inductive inference by finite automata

2008

AbstractFreivalds and Smith [R. Freivalds, C.H. Smith Memory limited inductive inference machines, Springer Lecture Notes in Computer Science 621 (1992) 19–29] proved that probabilistic limited memory inductive inference machines can learn with probability 1 certain classes of total recursive functions, which cannot be learned by deterministic limited memory inductive inference machines. We introduce quantum limited memory inductive inference machines as quantum finite automata acting as inductive inference machines. These machines, we show, can learn classes of total recursive functions not learnable by any deterministic, nor even by probabilistic, limited memory inductive inference machin…

Finite-state machineGeneral Computer Sciencebusiness.industryProbabilistic logicInductive inferenceInductive reasoningAutomataTheoretical Computer ScienceAutomatonTheoryofComputation_MATHEMATICALLOGICANDFORMALLANGUAGESQuantum computationLearningQuantum finite automataProbability distributionArtificial intelligencebusinessQuantumComputer Science(all)Quantum computerMathematicsTheoretical Computer Science
researchProduct

On the relative sizes of learnable sets

1998

Abstract Measure and category (or rather, their recursion-theoretical counterparts) have been used in theoretical computer science to make precise the intuitive notion “for most of the recursive sets”. We use the notions of effective measure and category to discuss the relative sizes of inferrible sets, and their complements. We find that inferable sets become large rather quickly in the standard hierarchies of learnability. On the other hand, the complements of the learnable sets are all large.

General Computer Science0102 computer and information sciencesMachine learningcomputer.software_genre01 natural sciencesMeasure (mathematics)Theoretical Computer ScienceTuring machinesymbols.namesake0101 mathematicsMathematicsBinary treeLearnabilitybusiness.industry010102 general mathematicsInductive inferenceCategoryInductive reasoningMeasureAbstract machine010201 computation theory & mathematicssymbolsArtificial intelligencebusinesscomputerComputer Science(all)Theoretical Computer Science
researchProduct

Inductive inference of recursive functions: complexity bounds

1991

This survey includes principal results on complexity of inductive inference for recursively enumerable classes of total recursive functions. Inductive inference is a process to find an algorithm from sample computations. In the case when the given class of functions is recursively enumerable it is easy to define a natural complexity measure for the inductive inference, namely, the worst-case mindchange number for the first n functions in the given class. Surely, the complexity depends not only on the class, but also on the numbering, i.e. which function is the first, which one is the second, etc. It turns out that, if the result of inference is Goedel number, then complexity of inference ma…

deterministicTheoryofComputation_MATHEMATICALLOGICANDFORMALLANGUAGESinductive inferencecomplexity boundspredictioncomplexityprobabilistic
researchProduct

Comparing various concepts of function prediction. Part 2.

1975

Prediction: f(m+1) is guessed from given f(0), ..., f(m). Program synthesis: a program computing f is guessed from given f(0), ..., f(m). The hypotheses are required to be correct for all sufficiently large m, or with some positive frequency. These approaches yield a hierarchy of function prediction and program synthesis concepts. The comparison problem of the concepts is solved.

deterministicfunction prediction:MATHEMATICS [Research Subject Categories]inductive inferenceprogram synthesis
researchProduct

Comparing various types of limiting synthesis and prediction of functions

1974

deterministicinductive inferenceprediction
researchProduct

Comparing various concepts of function prediction. Part 1.

1974

Prediction: f(m+1) is guessed from given f(0), ..., f(m). Program synthesis: a program computing f is guessed from given f(0), ..., f(m). The hypotheses are required to be correct for all sufficiently large m, or with some positive frequency. These approaches yield a hierarchy of function prediction and program synthesis concepts. The comparison problem of the concepts is solved.

function prediction:MATHEMATICS [Research Subject Categories]inductive inferenceprogram synthesis
researchProduct